14 research outputs found
Power Allocation and Cooperative Diversity in Two-Way Non-Regenerative Cognitive Radio Networks
In this paper, we investigate the performance of a dual-hop block fading
cognitive radio network with underlay spectrum sharing over independent but not
necessarily identically distributed (i.n.i.d.) Nakagami- fading channels.
The primary network consists of a source and a destination. Depending on
whether the secondary network which consists of two source nodes have a single
relay for cooperation or multiple relays thereby employs opportunistic relay
selection for cooperation and whether the two source nodes suffer from the
primary users' (PU) interference, two cases are considered in this paper, which
are referred to as Scenario (a) and Scenario (b), respectively. For the
considered underlay spectrum sharing, the transmit power constraint of the
proposed system is adjusted by interference limit on the primary network and
the interference imposed by primary user (PU). The developed new analysis
obtains new analytical results for the outage capacity (OC) and average symbol
error probability (ASEP). In particular, for Scenario (a), tight lower bounds
on the OC and ASEP of the secondary network are derived in closed-form. In
addition, a closed from expression for the end-to-end OC of Scenario (a) is
achieved. With regards to Scenario (b), a tight lower bound on the OC of the
secondary network is derived in closed-form. All analytical results are
corroborated using Monte Carlo simulation method
Rethinking Data Heterogeneity in Federated Learning: Introducing a New Notion and Standard Benchmarks
Though successful, federated learning presents new challenges for machine
learning, especially when the issue of data heterogeneity, also known as
Non-IID data, arises. To cope with the statistical heterogeneity, previous
works incorporated a proximal term in local optimization or modified the model
aggregation scheme at the server side or advocated clustered federated learning
approaches where the central server groups agent population into clusters with
jointly trainable data distributions to take the advantage of a certain level
of personalization. While effective, they lack a deep elaboration on what kind
of data heterogeneity and how the data heterogeneity impacts the accuracy
performance of the participating clients. In contrast to many of the prior
federated learning approaches, we demonstrate not only the issue of data
heterogeneity in current setups is not necessarily a problem but also in fact
it can be beneficial for the FL participants. Our observations are intuitive:
(1) Dissimilar labels of clients (label skew) are not necessarily considered
data heterogeneity, and (2) the principal angle between the agents' data
subspaces spanned by their corresponding principal vectors of data is a better
estimate of the data heterogeneity. Our code is available at
https://github.com/MMorafah/FL-SC-NIID.Comment: arXiv admin note: text overlap with arXiv:2209.1052
CEFHRI: A Communication Efficient Federated Learning Framework for Recognizing Industrial Human-Robot Interaction
Human-robot interaction (HRI) is a rapidly growing field that encompasses
social and industrial applications. Machine learning plays a vital role in
industrial HRI by enhancing the adaptability and autonomy of robots in complex
environments. However, data privacy is a crucial concern in the interaction
between humans and robots, as companies need to protect sensitive data while
machine learning algorithms require access to large datasets. Federated
Learning (FL) offers a solution by enabling the distributed training of models
without sharing raw data. Despite extensive research on Federated learning (FL)
for tasks such as natural language processing (NLP) and image classification,
the question of how to use FL for HRI remains an open research problem. The
traditional FL approach involves transmitting large neural network parameter
matrices between the server and clients, which can lead to high communication
costs and often becomes a bottleneck in FL. This paper proposes a
communication-efficient FL framework for human-robot interaction (CEFHRI) to
address the challenges of data heterogeneity and communication costs. The
framework leverages pre-trained models and introduces a trainable
spatiotemporal adapter for video understanding tasks in HRI. Experimental
results on three human-robot interaction benchmark datasets: HRI30, InHARD, and
COIN demonstrate the superiority of CEFHRI over full fine-tuning in terms of
communication costs. The proposed methodology provides a secure and efficient
approach to HRI federated learning, particularly in industrial environments
with data privacy concerns and limited communication bandwidth. Our code is
available at
https://github.com/umarkhalidAI/CEFHRI-Efficient-Federated-Learning.Comment: Accepted in IROS 202
When Do Curricula Work in Federated Learning?
An oft-cited open problem of federated learning is the existence of data
heterogeneity at the clients. One pathway to understanding the drastic accuracy
drop in federated learning is by scrutinizing the behavior of the clients' deep
models on data with different levels of "difficulty", which has been left
unaddressed. In this paper, we investigate a different and rarely studied
dimension of FL: ordered learning. Specifically, we aim to investigate how
ordered learning principles can contribute to alleviating the heterogeneity
effects in FL. We present theoretical analysis and conduct extensive empirical
studies on the efficacy of orderings spanning three kinds of learning:
curriculum, anti-curriculum, and random curriculum. We find that curriculum
learning largely alleviates non-IIDness. Interestingly, the more disparate the
data distributions across clients the more they benefit from ordered learning.
We provide analysis explaining this phenomenon, specifically indicating how
curriculum training appears to make the objective landscape progressively less
convex, suggesting fast converging iterations at the beginning of the training
procedure. We derive quantitative results of convergence for both convex and
nonconvex objectives by modeling the curriculum training on federated devices
as local SGD with locally biased stochastic gradients. Also, inspired by
ordered learning, we propose a novel client selection technique that benefits
from the real-world disparity in the clients. Our proposed approach to client
selection has a synergic effect when applied together with ordered learning in
FL
Recommended from our members
On Distributed Learning Techniques for Machine Learning
Modern machine and deep learning algorithms require a lot of data and computation and benefit from large and representative datasets (e.g., ImageNet and COCO), as much data and computational resources as possible should be gathered. However, in some of the applications, collecting abundant examples for certain classes in domains such as biology and medicine may be impossible in practice. For instance, in dermatology, there are some rare diseases with a small number of patients. It is, therefore, natural to ask can we train a model using data that is naturally dispersed among different parties in practice (e.g., edge devices and hospitals, etc.) without explicitly sharing their data? Federated Learning (FL) is a recently proposed distributed training framework in edge computing environment that enables distributed edge devices to collaboratively train a global model under the orchestration of a central server without compromising the privacy of their data. While FL has great potential, it faces challenges in practical settings, including statistical data heterogeneity (Non-IID data), personalization, fairness,
computation overhead, and communication cost. I designed techniques that could alleviate the mentioned challenges in FL. FL is explicitly designed for Non-IID edge devices. A global model can perform well on personalized predictions if the edge device’s context and personal data are nicely featured and embodied in the dataset, which is not the case for most edge devices. Most techniques for personalization either fail to build a model with low generalization error or are not very effective, especially when local distributions are far from the average distribution. In addition, when the edge devices are the edge devices, computational efficiency, and communication cost will be a crucial bottleneck, as edge devices are typically constrained by computational limitations and upload bandwidth of 1 MB/s or less.On the other side of the spectrum, due to redundancy, datasets' information content is much smaller than their actual volume, despite their steady growth. Existing techniques are not effective in identifying and extracting the non-redundant information content, considering the intrinsic structure of the data. Hence, machine learning models are trained on massive data volumes, which necessitates exceptionally large and expensive computational resources. A crucial challenge of machine learning today is to develop methods that can extract representative information volumes and accurately and robustly learn from the extracted representatives. My methods have immediate application to high-impact problems where massive data prohibits efficient learning and inference, such as GAN, recommender systems, graphs, video, and other large data streams. My research approach to uniquely address this challenge consists of (1) extracting the information volume by summarizing the most representative subsets. 2) At the core of my research lie rigorous and practical techniques that provide strong guarantees for the quality of the extracted representatives and the learned models' accuracy. My proposed methods open up new avenues for learning from representative data extracted from massive datasets
FLIS: Clustered Federated Learning Via Inference Similarity for Non-IID Data Distribution
Conventional federated learning (FL) approaches are ineffective in scenarios where clients have significant differences in the distributions of their local data. The Non-IID data distribution in the client data causes a drift in the local model updates from the global optima, which significantly impacts the performance of the trained models. In this article, we present a new algorithm called FLIS that aims to address this problem by grouping clients into clusters that have jointly trainable data distributions. This is achieved by comparing the inference similarity of client models. Our proposed framework captures settings where different groups of users may have their own objectives (learning tasks), but by aggregating their data with others in the same cluster (same learning task), superior models can be derived via more efficient and personalized federated learning. We present experimental results to demonstrate the benefits of FLIS over the state-of-the-art approaches on the CIFAR-100/10, SVHN, and FMNIST datasets. Our code is available at https://github.com/MMorafah/FLIS